Acta Psychologica Sinica ›› 2026, Vol. 58 ›› Issue (7): 1237-1253.doi: 10.3724/SP.J.1041.2026.1237
• Academic Papers of the 28th Annual Meeting of the China Association for Science and Technology • Next Articles
LI Chang-Jin1,2,3, JIAO Liying4, CHEN Zhen1,2,3, XU Hengbin1,2,3, WU Michael Shengtao5, XU Yan1,2,3
Received:2025-05-22
Published:2026-07-25
Online:2026-05-15
LI Chang-Jin, JIAO Liying, CHEN Zhen, XU Hengbin, WU Michael Shengtao, XU Yan. (2026). Personalized alignment of large language models and its impact on moral judgment. Acta Psychologica Sinica, 58(7), 1237-1253.
Add to citation manager EndNote|Ris|BibTeX
URL: https://journal.psych.ac.cn/acps/EN/10.3724/SP.J.1041.2026.1237
| [1] Ashton, M. C., & Lee, K. (2007). Empirical, theoretical, and practical advantages of the HEXACO model of personality structure. [2] Ashton, M. C., & Lee, K. (2008a). The HEXACO model of personality structure and the importance of the H Factor. [3] Ashton, M. C., & Lee, K. (2008b). The prediction of Honesty-Humility-related criteria by the HEXACO and Five-Factor Models of personality. [4] Ashton, M. C., & Lee, K. (2009). The HEXACO-60: A short measure of the major dimensions of personality. [5] Bender E. M., Gebru T., McMillan-Major A., & Shmitchell S. (2021). On the dangers of stochastic parrots: Can language models be too big?. In [6] Bodroža B., Dinić B. M., & Bojić L. (2024). Personality testing of large language models: Limited temporal stability, but highlighted prosociality. [7] Bonnefon J. F., Rahwan I., & Shariff A. (2024). The moral psychology of artificial intelligence. [8] Borman H., Leontjeva A., Pizzato L., Jiang M. K., & Jermyn D. (2024). [9] Chen R., Arditi A., Sleight H., Evans O., & Lindsey J. (2025). [10] Corrêa N. K., Galvão C., Santos J. W., Del Pino C., Pinto E. P., Barbosa C., ... de Oliveira N. (2023). Worldwide AI ethics: A review of 200 guidelines and recommendations for AI Governance. [11] DeYoung C. G., Peterson J. B., & Higgins D. M. (2002). Higher-order factors of the Big Five predict conformity: Are there neuroses of health? [12] Djeriouat, H., & Trémolière, B. (2014). The Dark Triad of personality and utilitarian moral judgment: The mediating role of Honesty/Humility and Harm/Care. [13] Gabriel, I. (2020). Artificial intelligence, values, and alignment. [14] Graham J., Meindl P., Beall E., Johnson K. M., & Zhang L. (2016). Cultural differences in moral judgment and behavior, across and within societies. [15] Hadi M. U., Tashi Q. A., Qureshi R., Shah A., Muneer A., Irfan M., ... Shah M. (2025). [16] Hagendorff, T. (2024). Deception abilities emerged in large language models. [17] Hagendorff T., Dasgupta I., Binz M., Chan S. C., Lampinen A., Wang J. X., Akata Z., & Schulz E. (2023). [18] He, J., & Liu, J. (2025). [19] Hilbig B. E., Glöckner A., & Zettler I. (2014). Personality and prosocial behavior: linking basic traits and social value orientations. [20] Hu X., Li M., Li Y., Li K., & Yu F. (2026). Moral deficiency in AI decision-making: Underlying mechanisms and mitigation strategies. [胡小勇, 李穆峰, 李悦, 李凯, 喻丰. (2026). 人工智能决策的道德缺失效应及其机制与应对策略. [21] Hu X., Li M., Wang D., & Yu F. (2024). Reactions to immoral AI decisions: The moral deficit effect and its underlying mechanism. [胡小勇, 李穆峰, 王笛新, 喻丰. (2024). 人工智能决策的道德缺失效应及其机制. [22] Jiang G., Xu M., Zhu S. C., Han W., Zhang C., & Zhu Y. (2023). Evaluating and inducing personality in pre-trained language models. In A. Oh, T. Naumann, A. Globerson, K. Saenko, M. Hardt, & S. Levine (Eds.), [23] Jiang H., Zhang X., Cao X., Breazeal C., Roy D., & Kabbara J. (2024). PersonaLLM: Investigating the ability of large language models to express personality traits. In K. Duh, H. Gomez, & S. Bethard (Eds.), [24] Jiao L., Li C.-J., Chen Z., Xu H., & Xu Y. (2025). When AI “possesses” personality: Roles of good and evil personalities influence moral judgment in large language models. [焦丽颖, 李昌锦, 陈圳, 许恒彬, 许燕. (2025). 当AI“具有”人格:善恶人格角色对大语言模型道德判断的影响. [25] Jiao L., Yang Y., Xu Y., Gao S., & Zhang H. (2019). Good and evil in Chinese culture: Personality structure and connotation. [焦丽颖, 杨颖, 许燕, 高树青, 张和云. (2019). 中国人的善与恶:人格结构与内涵. [26] Jobin A., Ienca M., & Vayena E. (2019). The global landscape of AI ethics guidelines. [27] Kroneisen, M., & Heck, D. W. (2020). Interindividual differences in the sensitivity for consequences, moral norms, and preferences for inaction: Relating basic personality traits to the CNI model. [28] Kruglanski A. W., Szumowska E., Kopetz C. H., Vallerand R. J., & Pierro A. (2021). On the psychology of extremism: How motivational imbalance breeds intemperance. [29] Lee, K., & Ashton, M. C. (2004). Psychometric properties of the HEXACO personality inventory. [30] Lee, K., & Ashton, M. C. (2008). The HEXACO personality factors in the indigenous personality lexicons of English and 11 other languages. [31] Lee, K., & Ashton, M. C. (2014). The dark triad, the big five, and the HEXACO model. [32] Lee K., Ashton M. C., Wiltshire J., Bourdage J. S., Visser B. A., & Gallucci A. (2013). Sex, power, and money: Prediction from the Dark Triad and Honesty-Humility. [33] Lin B. Y., Ravichander A., Lu X., Dziri N., Sclar M., Chandu K., Bhagavatula C., & Choi Y. (2023). [34] Lomas, T. (2019). The roots of virtue: A cross-cultural lexical analysis. [35] Lotto L., Manfrinati A., & Sarlo M. (2014). A new set of moral dilemmas: Norms for moral acceptability, decision times, and emotional salience. [36] Lu J. G., Song L. L., & Zhang L. D. (2025). Cultural tendencies in generative AI. [37] Matei, M.-C., & Abrudan, M.-M. (2018). Are national cultures changing? Evidence from the World Values Survey. [38] Minaee S., Mikolov T., Nikzad N., Chenaghlu M., Socher R., Amatriain X., & Gao J. (2024). [39] Mittelstadt, B. (2019). Principles alone cannot guarantee ethical AI. [40] Moser C., Den Hond F., & Lindebaum D. (2022). Morality in the age of artificially intelligent algorithms. [41] Newsham, L., & Prince, D. (2025). Personality-driven decision making in LLM-based autonomous agents. In S. Das, A. Nowé(General Chairs), & Y. Vorobeychik (Program Chair), [42] Ng, A. Y., & Russell, S. J. (2000). Algorithms for inverse reinforcement learning. In P. Langley (Ed.), [43] Nighojkar A., Moydinboyev B., Duong M., & Licato J. (2025). [44] Niszczota P., Janczak M., & Misiak M. (2025). Large language models can replicate cross-cultural differences in personality. [45] OpenAI. (2023). [46] Ramezani, A., & Xu, Y. (2023). Knowledge of cultural moral norms in large language models. In A. Rogers, J. Boyd-Graber, & N. Okazaki (Eds.), [47] Saucier G., Kenner J., Iurino K., Bou Malham P., Chen Z., Thalmayer A. G., ... Altschul C. (2014). Cross-cultural differences in a global “survey of world views”. [48] Serapio-García G., Safdari M., Crepy C., Sun L., Fitz S., Romero P., ... Matarić M. (2025). A psychometric framework for evaluating and shaping personality traits in large language models. [49] Shanahan M., McDonell K., & Reynolds L. (2023). Role play with large language models. [50] Sorokovikova A., Fedorova N., Rezagholi S., & Yamshchikov I. P. (2024). [51] Strus, W., & Cieciuch, J. (2021). Higher-order factors of the big six-similarities between big twos identified above the big five and the big six. [52] Thielmann I., Spadaro G., & Balliet D. (2020). Personality and prosocial behavior: A theoretical framework and meta-analysis. [53] Tong H., Lu E., Sun Y., Han Z., Liu C., Zhao F., & Zeng Y. (2024). [54] Treglown, L., & Furnham, A. (2026). AI, social desirability, and personality assessments: Impression management in large language models. [55] Wang P., Zou H., Yan Z., Guo F., Sun T., Xiao Z., & Zhang B. (2024). [56] Wang S., Li R., Chen X., Yuan Y., Yang M., & Wong D. F. (2025). Exploring the impact of personality traits on LLM bias and toxicity. In C. Christodoulopoulos, T. Chakraborty, C. Rose, & V. Peng (Eds.), [57] Wang X., Duan S., Yi X., Yao J., Zhou S., Wei Z., ... Xie X. (2024). On the essence and prospect: An investigation of alignment approaches for big models. In K. Larson (Ed.), [58] Wu, M. S., & Peng, K. (2025). Human advantages and psychological transformations in the era of artificial intelligence. [吴胜涛, 彭凯平. (2025). 智能时代的人类优势与心理变革(代序). [59] Xu, Y. (2024). Personality psychology (3rd ed.). Beijing, China: Beijing Normal University Publishing Group. [许燕. (2024). 人格心理学 (第3版). 北京: 北京师范大学出版社.] [60] Xu Z., Sengar N., Chen T., Chung H., & Oviedo- Trespalacios O. (2025). Where is morality on wheels? Decoding large language model (LLM)-driven decision in the ethical dilemmas of autonomous vehicles. [61] Yao J., Yi X., Wang X., Wang J., & Xie X. (2023). [62] Yu, B., & Kim, J. (2024). Personality of AI. In L. Rutkowski, R. Scherer, M. Korytkowski, W. Pedrycz, R. Tadeusiewicz, & J. M. Zurada (Eds.), [63] Yuan X., Hu J., & Zhang Q. (2024). [64] Zaim bin Ahmad, M. S., & Takemoto, K. (2025). Large-scale moral machine experiment on large language models. [65] Zhao, G. L. (2023-06-20). Actual scores surpass ChatGPT! Baidu Wenxin Large Model Version 3.5 internal test application. [赵广立. (2023-06-20). 实测得分超ChatGPT!百度文心大模型3.5版内测应用. [66] Zhou C., Liu P., Xu P., Iyer S., Sun J., Mao Y., ... Levy O. (2023). LIMA: less is more for alignment. In A. Oh, T. Naumann, A. Globerson, K. Saenko, M. Hardt, & S. Levine (Eds.), [67] Zhou, X., & Liu, H. (2024). New ethical challenges in the digital and intelligent era. [周欣悦, 刘惠洁. (2024). 数智时代面临新的伦理挑战(前言). |
| Viewed | ||||||
|
Full text |
|
|||||
|
Abstract |
|
|||||